Distributed training and scalability for the particle clustering method UCluster
نویسندگان
چکیده
In recent years, machine-learning methods have become increasingly important for the experiments at Large Hadron Collider (LHC). They are utilised in everything from trigger systems to reconstruction and data analysis. The UCluster method is a general model providing unsupervised clustering of particle physics data, that can be easily modified provide solutions variety different decision problems. current paper, we improve on by adding option training scalable distributed fashion, thereby extending its utility learn arbitrarily large sets. combines graph-based neural network called ABCnet with step, using combined loss function phase. original code publicly available TensorFlow v1.14 has previously been trained single GPU. It shows accuracy 81% when applied problem multi-class classification simulated jet events. Our implementation adds functionality utilising Horovod framework, which necessitated migration v2. Together parquet files splitting up between compute nodes, makes any amount input something will essential use real LHC We find well suited training, time decreasing direct relation number GPU's used. However, further improvements more exhaustive possibly hyper-parameter search required order achieve reported method.
منابع مشابه
Entropy-based Consensus for Distributed Data Clustering
The increasingly larger scale of available data and the more restrictive concerns on their privacy are some of the challenging aspects of data mining today. In this paper, Entropy-based Consensus on Cluster Centers (EC3) is introduced for clustering in distributed systems with a consideration for confidentiality of data; i.e. it is the negotiations among local cluster centers that are used in t...
متن کاملScalability Considerations and Evaluation for Distributed Mission Training Systems
Distributed Mission Training is a promising technology in support of training in general and within DoD in particular. Architectural considerations (interplay of hardware and software) is instrumental given high demands for network bandwidth and computing resources in relation to proliferating software complexity. This paper proposes a co-design modeling and simulation methodology and environme...
متن کاملMultiobjective Hboa, Clustering, and Scalability Multiobjective Hboa, Clustering, and Scalability
This paper describes a scalable algorithm for solving multiobjective decomposable problems by combining the hierarchical Bayesian optimization algorithm (hBOA) with the nondominated sorting genetic algorithm (NSGA-II) and clustering in the objective space. It is first argued that for good scalability, clustering or some other form of niching in the objective space is necessary and the size of e...
متن کاملthe search for the self in becketts theatre: waiting for godot and endgame
this thesis is based upon the works of samuel beckett. one of the greatest writers of contemporary literature. here, i have tried to focus on one of the main themes in becketts works: the search for the real "me" or the real self, which is not only a problem to be solved for beckett man but also for each of us. i have tried to show becketts techniques in approaching this unattainable goal, base...
15 صفحه اولParticle swarm optimization method for image clustering
An image clustering method that is based on the particle swarm optimizer (PSO) is developed in this paper. The algorithm finds the centroids of a user specified number of clusters, where each cluster groups together similar image primitives. To illustrate its wide applicability, the proposed image classifier has been applied to synthetic, MRI and satellite images. Experimental results show that...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Epj Web of Conferences
سال: 2021
ISSN: ['2101-6275', '2100-014X']
DOI: https://doi.org/10.1051/epjconf/202125102054